Goto

Collaborating Authors

 transferable normalization


Transferable Normalization: Towards Improving Transferability of Deep Neural Networks

Neural Information Processing Systems

Deep neural networks (DNNs) excel at learning representations when trained on large-scale datasets. Pre-trained DNNs also show strong transferability when fine-tuned to other labeled datasets. However, such transferability becomes weak when the target dataset is fully unlabeled as in Unsupervised Domain Adaptation (UDA). We envision that the loss of transferability may stem from the intrinsic limitation of the architecture design of DNNs. In this paper, we delve into the components of DNN architectures and propose Transferable Normalization (TransNorm) in place of existing normalization techniques. TransNorm is an end-to-end trainable layer to make DNNs more transferable across domains. As a general method, TransNorm can be easily applied to various deep neural networks and domain adaption methods, without introducing any extra hyper-parameters or learnable parameters. Empirical results justify that TransNorm not only improves classification accuracies but also accelerates convergence for mainstream DNN-based domain adaptation methods.



Transferable Normalization: Towards Improving Transferability of Deep Neural Networks

Neural Information Processing Systems

Deep neural networks (DNNs) excel at learning representations when trained on large-scale datasets. Pre-trained DNNs also show strong transferability when fine-tuned to other labeled datasets. However, such transferability becomes weak when the target dataset is fully unlabeled as in Unsupervised Domain Adaptation (UDA). We envision that the loss of transferability may stem from the intrinsic limitation of the architecture design of DNNs. In this paper, we delve into the components of DNN architectures and propose Transferable Normalization (TransNorm) in place of existing normalization techniques.


Reviews: Transferable Normalization: Towards Improving Transferability of Deep Neural Networks

Neural Information Processing Systems

This paper is different from most works focus on reducing the domain shift from perspective of loss functions, contributing to network design by developing a novel transferable normalization (TranNorm) layer. TranNorm is well motivated, separately normalizing source and target features in a minibatch and meanwhile weighting each channel in terms of transferability. It is clear different from and meanwhile significantly outperformes related methods, e.g., AdaBN [15] and AutoDIAL [21]. The TranNorm layer is simple and free of parameters, which can be conveniently plugged in mainstream networks. I think that this work will have a non-trivial impact: the proposed TranNorm can be used as backbone layer improving other state-of-the-art methods.


Reviews: Transferable Normalization: Towards Improving Transferability of Deep Neural Networks

Neural Information Processing Systems

The paper proposes a novel normalization strategies in the line of AdaBN and AutoDIAL paper. The method is simple, easy to use and shows better results than other normalization strategies. The paper is well written, the experimental evaluation is large and solid with the help of the complementary information provided in the rebuttal.


Transferable Normalization: Towards Improving Transferability of Deep Neural Networks

Neural Information Processing Systems

Deep neural networks (DNNs) excel at learning representations when trained on large-scale datasets. Pre-trained DNNs also show strong transferability when fine-tuned to other labeled datasets. However, such transferability becomes weak when the target dataset is fully unlabeled as in Unsupervised Domain Adaptation (UDA). We envision that the loss of transferability may stem from the intrinsic limitation of the architecture design of DNNs. In this paper, we delve into the components of DNN architectures and propose Transferable Normalization (TransNorm) in place of existing normalization techniques.


Transferable Normalization: Towards Improving Transferability of Deep Neural Networks

Wang, Ximei, Jin, Ying, Long, Mingsheng, Wang, Jianmin, Jordan, Michael I.

Neural Information Processing Systems

Deep neural networks (DNNs) excel at learning representations when trained on large-scale datasets. Pre-trained DNNs also show strong transferability when fine-tuned to other labeled datasets. However, such transferability becomes weak when the target dataset is fully unlabeled as in Unsupervised Domain Adaptation (UDA). We envision that the loss of transferability may stem from the intrinsic limitation of the architecture design of DNNs. In this paper, we delve into the components of DNN architectures and propose Transferable Normalization (TransNorm) in place of existing normalization techniques.